gender recognition
Exploring the Feasibility of Deep Learning Techniques for Accurate Gender Classification from Eye Images
Hasan, Basna Mohammed Salih, Mstafa, Ramadhan J.
Gender classification has emerged as a crucial aspect in various fields, including security, human-machine interaction, surveillance, and advertising. Nonetheless, the accuracy of this classification can be influenced by factors such as cosmetics and disguise. Consequently, our study is dedicated to addressing this concern by concentrating on gender classification using color images of the periocular region. The periocular region refers to the area surrounding the eye, including the eyelids, eyebrows, and the region between them. It contains valuable visual cues that can be used to extract key features for gender classification. This paper introduces a sophisticated Convolutional Neural Network (CNN) model that utilizes color image databases to evaluate the effectiveness of the periocular region for gender classification. To validate the model's performance, we conducted tests on two eye datasets, namely CVBL and (Female and Male). The recommended architecture achieved an outstanding accuracy of 99% on the previously unused CVBL dataset while attaining a commendable accuracy of 96% with a small number of learnable parameters (7,235,089) on the (Female and Male) dataset. To ascertain the effectiveness of our proposed model for gender classification using the periocular region, we evaluated its performance through an extensive range of metrics and compared it with other state-of-the-art approaches. The results unequivocally demonstrate the efficacy of our model, thereby suggesting its potential for practical application in domains such as security and surveillance.
- South America > Argentina > Patagonia > Río Negro Province > Viedma (0.05)
- Asia > Middle East > Iraq > Kurdistan Region > Duhok Governorate > Zakho (0.05)
- Asia > Middle East > Iraq > Kurdistan Region > Duhok Governorate > Duhok (0.04)
- Europe > Netherlands (0.04)
Beyond the binary: Limitations and possibilities of gender-related speech technology research
Sanchez, Ariadna, Ross, Alice, Markl, Nina
This paper presents a review of 107 research papers relating to speech and sex or gender in ISCA Interspeech publications between 2013 and 2023. We note the scarcity of work on this topic and find that terminology, particularly the word gender, is used in ways that are underspecified and often out of step with the prevailing view in social sciences that gender is socially constructed and is a spectrum as opposed to a binary category. We draw attention to the potential problems that this can cause for already marginalised groups, and suggest some questions for researchers to ask themselves when undertaking work on speech and gender.
- Oceania > New Zealand (0.04)
- Oceania > Australia (0.04)
- Europe > United Kingdom > England > Essex (0.04)
- (4 more...)
- Research Report (1.00)
- Overview (0.88)
- Health & Medicine (1.00)
- Law > Civil Rights & Constitutional Law (0.68)
Acoustic models of Brazilian Portuguese Speech based on Neural Transformers
Gauy, Marcelo Matheus, Finger, Marcelo
An acoustic model, trained on a significant amount of unlabeled data, consists of a self-supervised learned speech representation useful for solving downstream tasks, perhaps after a fine-tuning of the model in the respective downstream task. In this work, we build an acoustic model of Brazilian Portuguese Speech through a Transformer neural network. This model was pretrained on more than $800$ hours of Brazilian Portuguese Speech, using a combination of pretraining techniques. Using a labeled dataset collected for the detection of respiratory insufficiency in Brazilian Portuguese speakers, we fine-tune the pretrained Transformer neural network on the following tasks: respiratory insufficiency detection, gender recognition and age group classification. We compare the performance of pretrained Transformers on these tasks with that of Transformers without previous pretraining, noting a significant improvement. In particular, the performance of respiratory insufficiency detection obtains the best reported results so far, indicating this kind of acoustic model as a promising tool for speech-as-biomarker approach. Moreover, the performance of gender recognition is comparable to the state of the art models in English.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- South America > Brazil > Pernambuco > Recife (0.04)
- South America > Brazil > Rio Grande do Sul > Porto Alegre (0.04)
- (4 more...)
MiVOLO: Multi-input Transformer for Age and Gender Estimation
Kuprashevich, Maksim, Tolstykh, Irina
Age and gender recognition in the wild is a highly challenging task: apart from the variability of conditions, pose complexities, and varying image quality, there are cases where the face is partially or completely occluded. We present MiVOLO (Multi Input VOLO), a straightforward approach for age and gender estimation using the latest vision transformer. Our method integrates both tasks into a unified dual input/output model, leveraging not only facial information but also person image data. This improves the generalization ability of our model and enables it to deliver satisfactory results even when the face is not visible in the image. To evaluate our proposed model, we conduct experiments on four popular benchmarks and achieve state-of-the-art performance, while demonstrating real-time processing capabilities. Additionally, we introduce a novel benchmark based on images from the Open Images Dataset. The ground truth annotations for this benchmark have been meticulously generated by human annotators, resulting in high accuracy answers due to the smart aggregation of votes. Furthermore, we compare our model's age recognition performance with human-level accuracy and demonstrate that it significantly outperforms humans across a majority of age ranges. Finally, we grant public access to our models, along with the code for validation and inference. In addition, we provide extra annotations for used datasets and introduce our new benchmark.
Manipulating Transfer Learning for Property Inference
Tian, Yulong, Suya, Fnu, Suri, Anshuman, Xu, Fengyuan, Evans, David
Transfer learning is a popular method for tuning pretrained (upstream) models for different downstream tasks using limited data and computational resources. We study how an adversary with control over an upstream model used in transfer learning can conduct property inference attacks on a victim's tuned downstream model. For example, to infer the presence of images of a specific individual in the downstream training set. We demonstrate attacks in which an adversary can manipulate the upstream model to conduct highly effective and specific property inference attacks (AUC score $> 0.9$), without incurring significant performance loss on the main task. The main idea of the manipulation is to make the upstream model generate activations (intermediate features) with different distributions for samples with and without a target property, thus enabling the adversary to distinguish easily between downstream models trained with and without training examples that have the target property. Our code is available at https://github.com/yulongt23/Transfer-Inference.
SVLDL: Improved Speaker Age Estimation Using Selective Variance Label Distribution Learning
Kang, Zuheng, Wang, Jianzong, Peng, Junqing, Xiao, Jing
Estimating age from a single speech is a classic and challenging topic. Although Label Distribution Learning (LDL) can represent adjacent indistinguishable ages well, the uncertainty of the age estimate for each utterance varies from person to person, i.e., the variance of the age distribution is different. To address this issue, we propose selective variance label distribution learning (SVLDL) method to adapt the variance of different age distributions. Furthermore, the model uses WavLM as the speech feature extractor and adds the auxiliary task of gender recognition to further improve the performance. Two tricks are applied on the loss function to enhance the robustness of the age estimation and improve the quality of the fitted age distribution. Extensive experiments show that the model achieves state-of-the-art performance on all aspects of the NIST SRE08-10 and a real-world datasets.
- Asia > China > Guangdong Province > Shenzhen (0.04)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
Is your smile male or female? Mapping the dynamics of a smile to enable gender recognition
Although automatic gender recognition is already available, existing methods use static images and compare fixed facial features. The new research, by the University of Bradford, is the first to use the dynamic movement of the smile to automatically distinguish between men and women. Led by Professor Hassan Ugail, the team mapped 49 landmarks on the face, mainly around the eyes, mouth and down the nose. They used these to assess how the face changes as we smile caused by the underlying muscle movements -- including both changes in distances between the different points and the'flow' of the smile: how much, how far and how fast the different points on the face moved as the smile was formed. They then tested whether there were noticeable differences between men and women -- and found that there were, with women's smiles being more expansive.
AI can now tell if you're a man or a woman, just by your smile
Men and women have different patterns of smiling, new research reports -- and this, the authors add, can allow AI to easily distinguish between the genders. Image credits Benjamin D. Glass / U.S. Navy. Many a man has been enraptured by the right smile, and many more will probably follow -- although the opposite doesn't seem to hold true. Regardless, while romance unfolds across the world, one team of researchers from the University of Bradford is working to bring this subtle yet powerful gesture to bear in our interactions with artificial intelligence (AI). According to them, computers can learn to differentiate between men or women simply by observing a smile. Led by Professor Hassan Ugail, the team mapped 49 distinct points (or'landmarks) on smiling human faces -- mainly around the eyes, mouth, and down the nose.
Understanding and Comparing Deep Neural Networks for Age and Gender Classification
Lapuschkin, Sebastian, Binder, Alexander, Müller, Klaus-Robert, Samek, Wojciech
Recently, deep neural networks have demonstrated excellent performances in recognizing the age and gender on human face images. However, these models were applied in a black-box manner with no information provided about which facial features are actually used for prediction and how these features depend on image preprocessing, model initialization and architecture choice. We present a study investigating these different effects. In detail, our work compares four popular neural network architectures, studies the effect of pretraining, evaluates the robustness of the considered alignment preprocessings via cross-method test set swapping and intuitively visualizes the model's prediction strategies in given preprocessing conditions using the recent Layer-wise Relevance Propagation (LRP) algorithm. Our evaluations on the challenging Adience benchmark show that suitable parameter initialization leads to a holistic perception of the input, compensating artefactual data representations. With a combination of simple preprocessing steps, we reach state of the art performance in gender recognition.
- Europe > Germany > Berlin (0.04)
- Asia > Singapore (0.04)
- North America > United States > Texas (0.04)
- North America > United States > Massachusetts > Hampshire County > Amherst (0.04)
Gender Recognition Based on Sift Features
Yousefi, Sahar, Zahedi, Morteza
This paper proposes a robust approach for face detection and gender classification in color images. Previous researches about gender recognition suppose an expensive computational and time-consuming pre-processing step in order to alignment in which face images are aligned so that facial landmarks like eyes, nose, lips, chin are placed in uniform locations in image. In this paper, a novel technique based on mathematical analysis is represented in three stages that eliminates alignment step. First, a new color based face detection method is represented with a better result and more robustness in complex backgrounds. Next, the features which are invariant to affine transformations are extracted from each face using scale invariant feature transform (SIFT) method. To evaluate the performance of the proposed algorithm, experiments have been conducted by employing a SVM classifier on a database of face images which contains 500 images from distinct people with equal ratio of male and female.
- North America > Canada (0.04)
- Asia > Middle East > Iran (0.04)